Welcome![Sign In][Sign Up]
Location:
Search - backpropagation neural network matlab

Search list

[AI-NN-PRbpexamples

Description: 使用MATLAB6.5实现的自适应学习速率反向传播网络-Matlab6.5 achieved using adaptive learning rate back-propagation network
Platform: | Size: 8192 | Author: 陈雷 | Hits:

[AI-NN-PRmybp(Jo)

Description: 我做的一个用matlab程序编写的BP算法,是是我毕设的一部分,程序运行绝对没有问题,欢迎大家指正。-I do a Matlab programming with the back-propagation algorithm, is part of a complete, running absolutely no problem, we are happy to correct.
Platform: | Size: 3072 | Author: FX | Hits:

[AI-NN-PRrbfn

Description: 利用MATLAB对神经网络进行编程,用newff()创建两层前向网络。网络输入范围[-1 1],第一层有10个tansig神经元-using MATLAB right neural network programming with newff () to the creation of a two-tier network. Network input range [-1 1], the first layer 10 tansig neurons
Platform: | Size: 5120 | Author: 龙海侠 | Hits:

[AI-NN-PRAdaptive

Description: The adaptive Neural Network Library is a collection of blocks that implement several Adaptive Neural Networks featuring different adaptation algorithms.~..~ There are 11 blocks that implement basically these 5 kinds of neural networks: 1) Adaptive Linear Network (ADALINE) 2) Multilayer Layer Perceptron with Extended Backpropagation algorithm (EBPA) 3) Radial Basis Functions (RBF) Networks 4) RBF Networks with Extended Minimal Resource Allocating algorithm (EMRAN) 5) RBF and Piecewise Linear Networks with Dynamic Cell Structure (DCS) algorithm A simulink example regarding the approximation of a scalar nonlinear function of 4 variables is included-The adaptive Neural Network Library is a collection of blocks that implement several Adaptive Neural Networks featuring different adaptation algorithms .~..~ There are 11 blocks that implement basically these five kinds of neural networks : a) Adaptive Linear Network (ADALINE) 2) Multilayer Layer 102206 with Extended Backpropagation algorithm (EBPA) 3) Radial Basis Functions (RBF) Networks, 4) RBF Networks with Extended Minimal Resource Allocating algorithm (EMRAN) 5) and RBF Networks with Piecewise Linear Dynamic Cell Structure (DCS) algorithm A Simulink example regarding the approximation of a scalar nonlinear function of four variables is included
Platform: | Size: 198656 | Author: 叶建槐 | Hits:

[AI-NN-PRgabp-src

Description: a matlab code for training a backpropagation neural network
Platform: | Size: 257024 | Author: ssss | Hits:

[Software Engineeringbackpropagation

Description: backpropagation mlp neural network
Platform: | Size: 1024 | Author: niksu | Hits:

[Education soft systemBackpropagation

Description: Make a program to train a Backpropagation neural network which simulates the behavior of an XOR gate.
Platform: | Size: 497664 | Author: Juan Carlos | Hits:

[Graph Recognizedctannprotected

Description: High information redundancy and correlation in face images result in efficiencies when such images are used directly for recognition. In this paper, discrete cosine transforms are used to reduce image information redundancy because only a subset of the transform coefficients are necessary to preserve the most important facial features such as hair outline, eyes and mouth. We demonstrate experimentally that when DCT coefficients are fed into a backpropagation neural network for classification, a high recognition rate can be achieved by using a very small proportion of transform coefficients. This makes DCT-based face recognition much faster than other approaches.-High information redundancy and correlation in face images result in inefficiencies when such images are used directly for recognition. In this paper, discrete cosine transforms are used to reduce image information redundancy because only a subset of the transform coefficients are necessary to preserve the most important facial features such as hair outline, eyes and mouth. We demonstrate experimentally that when DCT coefficients are fed into a backpropagation neural network for classification, a high recognition rate can be achieved by using a very small proportion of transform coefficients. This makes DCT-based face recognition much faster than other approaches.
Platform: | Size: 25600 | Author: mhm | Hits:

[matlabbpnn

Description: MATLAB neural network backprop code This code implements the basic backpropagation of error learning algorithm. The network has tanh hidden neurons and a linear output neuron. adjust the learning rate with the slider -MATLAB neural network backprop code This code implements the basic backpropagation of error learning algorithm. The network has tanh hidden neurons and a linear output neuron. adjust the learning rate with the slider
Platform: | Size: 2048 | Author: azoma | Hits:

[Education soft systemcharacterrecognition

Description: This the error backpropagation algorithm based character recognition program developed in matlab. The program folder contains the training images of some english alphabets & numerals. & trains the neural network using train command in matlab. The neural network is initialized using newff command. After the simulation, the program recognizes the input character. Accuracy on the provided images is 100 .-This is the error backpropagation algorithm based character recognition program developed in matlab. The program folder contains the training images of some english alphabets & numerals. & trains the neural network using train command in matlab. The neural network is initialized using newff command. After the simulation, the program recognizes the input character. Accuracy on the provided images is 100 .
Platform: | Size: 7730176 | Author: manish | Hits:

[AI-NN-PRFuzzy-Neural-Network-by-matlab

Description: 这是一个四个不同的S函数实现集合的递归模糊神经网络(RFNN)。该网络采用了4组可调参数,这使得它非常适合在线学习/操作,从而可应用到系统识别等方面。-This is a collection of four different S-function implementations of the recurrent fuzzy neural network (RFNN) described in detail in [1]. It is a four-layer, neuro-fuzzy network trained exclusively by error backpropagation at layers 2 and 4. The network employs 4 sets of adjustable parameters. In Layer 2: mean[i,j], sigma[i,j] and Theta[i,j] and in Layer 4: Weights w4[m,j]. The network uses considerably less adjustable parameters than ANFIS/CANFIS and therefore, its training is generally faster. This makes it ideal for on-line learning/operation. Also, its approximating/mapping power is increased due to the employment of dynamic elements within Layer 2. Scatter-type and Grid-type methods are selected for input space partitioning.
Platform: | Size: 117760 | Author: 林真天 | Hits:

[AI-NN-PRfirnet

Description: 有限脉冲神经网络MATLAB源码,由波兰国立大学生物医药信号处理实验室提供。FIR 神经网络属于一种BP神经网络。-Finite Impulse Response (FIR) neural network models each synapse as a linear filter to provide dynamic interconnectivity. Temporal backpropagation is used to train the network in which error terms are symmetrically filtered backward through the network
Platform: | Size: 382976 | Author: zhanglei | Hits:

[OtherMLP with BP

Description: neural network multilayer persptron trained by backpropagation
Platform: | Size: 201026 | Author: rafal@hussain | Hits:

[Graph Recognizespam_email

Description: 利用BP神经网络对邮件分类,2分类,实现反向传播算法细节,非直接函数调用(BP neural network is used to classify mail, 2 classifications, to implement the details of backpropagation algorithm, non direct function call.)
Platform: | Size: 2048 | Author: horizonch115 | Hits:

CodeBus www.codebus.net